Sequential Information Maximization: When is Greedy Near-optimal?
نویسندگان
چکیده
Optimal information gathering is a central challenge in machine learning and science in general. A common objective that quantifies the usefulness of observations is Shannon’s mutual information, defined w.r.t. a probabilistic model. Greedily selecting observations that maximize the mutual information is the method of choice in numerous applications, ranging from Bayesian experimental design to automated diagnosis, to active learning in Bayesian models. Despite its importance and widespread use in applications, little is known about the theoretical properties of sequential information maximization, in particular under noisy observations. In this paper, we analyze the widely used greedy policy for this task, and identify problem instances where it provides provably near-maximal utility, even in the challenging setting of persistent noise. Our results depend on a natural separability condition associated with a channel injecting noise into the observations. We also identify examples where this separability parameter is necessary in the bound: if it is too small, then the greedy policy fails to select informative tests.
منابع مشابه
Near-optimal Batch Mode Active Learning and Adaptive Submodular Optimization
Active learning can lead to a dramatic reduction in labeling effort. However, in many practical implementations (such as crowdsourcing, surveys, high-throughput experimental design), it is preferable to query labels for batches of examples to be labelled in parallel. While several heuristics have been proposed for batch-mode active learning, little is known about their theoretical performance. ...
متن کاملBayesian adaptive estimation under a random cost of observation associated with each observable variable ∗
In this paper, we adopt a decision theoretic view to Bayesian adaptive estimation. We extend the framework to situations where each observable variable is associated with a certain random cost of observation and consider the goal of maximizing the expected utility of a sequential experiment that ends when the total cost overruns a given budget. For example, the cost could be defined as the rand...
متن کاملAdaptive Maximization of Pointwise Submodular Functions With Budget Constraint
We study the worst-case adaptive optimization problem with budget constraint that is useful for modeling various practical applications in artificial intelligence and machine learning. We investigate the near-optimality of greedy algorithms for this problem with both modular and non-modular cost functions. In both cases, we prove that two simple greedy algorithms are not near-optimal but the be...
متن کاملOn Greedy Maximization of Entropy
Submodular function maximization is one of the key problems that arise in many machine learning tasks. Greedy selection algorithms are the proven choice to solve such problems, where prior theoretical work guarantees (1 − 1/e) approximation ratio. However, it has been empirically observed that greedy selection provides almost optimal solutions in practice. The main goal of this paper is to expl...
متن کاملConstrained Maximization of Non-Monotone Submodular Functions
The problem of constrained submodular maximization has long been studied, with near-optimal results known under a variety of constraints when the submodular function is monotone. The case of nonmonotone submodular maximization is not as well understood: the first approximation algorithms even for unconstrainted maximization were given by Feige et al. [FMV07]. More recently, Lee et al. [LMNS09] ...
متن کامل